CS168: The Modern Algorithmic Toolbox Lecture #10: Tensors, and Low-Rank Tensor Recovery

نویسندگان

  • Tim Roughgarden
  • Gregory Valiant
چکیده

Definition 1.1 A n1× n2× . . .× nk k-tensor is a set of n1 · n2 · . . . · nk numbers, which one interprets as being arranged in a k-dimensional hypercube. Given such a k-tensor, A, we can refer to a specific element via Ai1,i2,...,ik . A 2-tensor is simply a matrix, with Ai,j referring to the i, jth entry. You should think of a n1×n2×n3 3-tensor as simply a stack of n3 matrices, where each matrix has size n1×n2. The entry Ai,j,k of such a 3-tensor will refer to the i, jth entry of the kth matrix in the stack.

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

CS168: The Modern Algorithmic Toolbox Lecture #14: Linear and Convex Programming, with Applications to Sparse Recovery

Recall the setup in compressive sensing. There is an unknown signal z ∈ R, and we can only glean information about z through linear measurements. We choose m linear measurements a1, . . . , am ∈ R. “Nature” then chooses a signal z, and we receive the results b1 = 〈a1, z〉, . . . , bm = 〈am, z〉 of our measurements, when applied to z. The goal is then to recover z from b. Last lecture culminated i...

متن کامل

CS168: The Modern Algorithmic Toolbox Lecture #18: Linear and Convex Programming, with Applications to Sparse Recovery

Recall the setup in compressive sensing. There is an unknown signal z ∈ R, and we can only glean information about z through linear measurements. We choose m linear measurements a1, . . . , am ∈ R. “Nature” then chooses a signal z, and we receive the results b1 = 〈a1, z〉, . . . , bm = 〈am, z〉 of our measurements, when applied to z. The goal is then to recover z from b. Last lecture culminated i...

متن کامل

CS168: The Modern Algorithmic Toolbox Lecture #5: Sampling and Estimation

This week, we will cover tools for making inferences based on random samples drawn from some distribution of interest (e.g. a distribution over voter priorities, customer behavior, ip addresses, etc.). We will also learn how to use sampling techniques to solve hard problems— both problems that inherently involve randomness, as well as those that do not. As a warmup, to get into the probabilisti...

متن کامل

CS168: The Modern Algorithmic Toolbox Lecture #13: Sampling and Estimation

This week, we will cover tools for making inferences based on random samples drawn from some distribution of interest (e.g. a distribution over voter priorities, customer behavior, ip addresses, etc.). We will also learn how to use sampling techniques to solve hard problems— both problems that inherently involve randomness, as well as those that do not. As a warmup, to get into the probabilisti...

متن کامل

Optimal Low-Rank Tensor Recovery from Separable Measurements: Four Contractions Suffice

Tensors play a central role in many modern machine learning and signal processing applications. In such applications, the target tensor is usually of low rank, i.e., can be expressed as a sum of a small number of rank one tensors. This motivates us to consider the problem of low rank tensor recovery from a class of linear measurements called separable measurements. As specific examples, we focu...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 2015